Improved Iteratively Reweighted Least Squares for Unconstrained
نویسنده
چکیده
In this paper, we first study q minimization and its associated iterative reweighted algorithm for recovering sparse vectors. Unlike most existing work, we focus on unconstrained q minimization, for which we show a few advantages on noisy measurements and/or approximately sparse vectors. Inspired by the results in [Daubechies et al., Comm. Pure Appl. Math., 63 (2010), pp. 1–38] for constrained q minimization, we start with a preliminary yet novel analysis for unconstrained q minimization, which includes convergence, error bound, and local convergence behavior. Then, the algorithm and analysis are extended to the recovery of low-rank matrices. The algorithms for both vector and matrix recovery have been compared to some state-of-the-art algorithms and show superior performance on recovering sparse vectors and low-rank matrices.
منابع مشابه
Improved Iteratively Reweighted Least Squares for Unconstrained Smoothed 퓁q Minimization
In this paper, we first study q minimization and its associated iterative reweighted algorithm for recovering sparse vectors. Unlike most existing work, we focus on unconstrained q minimization, for which we show a few advantages on noisy measurements and/or approximately sparse vectors. Inspired by the results in [Daubechies et al., Comm. Pure Appl. Math., 63 (2010), pp. 1–38] for constrained ...
متن کاملConvergence Analysis of Generalized Iteratively Reweighted Least Squares Algorithms on Convex Function Spaces
The computation of robust regression estimates often relies on minimization of a convex functional on a convex set. In this paper we discuss a general technique for a large class of convex functionals to compute the minimizers iteratively which is closely related to majorization-minimization algorithms. Our approach is based on a quadratic approximation of the functional to be minimized and inc...
متن کاملA Parallel Min-Cut Algorithm using Iteratively Reweighted Least Squares
We present a parallel algorithm for the undirected s-t min-cut problem with floating-point valued weights. Our overarching algorithm uses an iteratively reweighted least squares framework. This generates a sequence of Laplacian linear systems, which we solve using parallel matrix algorithms. Our overall implementation is up to 30-times faster than a serial solver when using 128 cores.
متن کاملOn the Properties of Preconditioners for Robust Linear Regression
In this paper, we consider solving the robust linear regression problem, y = Ax+ ε by Newton’s method and iteratively reweighted least squares method. We show that each of these methods can be combined with preconditioned conjugate gradient least squares algorithm to solve large, sparse, rectangular systems of linear, algebraic equations efficiently. We consider the constant preconditioner A A ...
متن کاملIteratively Reweighted Least Squares for Maximum Likelihood Estimation, and some Robust and Resistant Alternatives
The scope of application of iteratively reweighted least squares to statistical estimation problems is considerably wider than is generally appreciated. It extends beyond the exponential-family-type generalized linear models to other distributions, to non-linear parameterizations, and to dependent observations. Various criteria for estimation other than maximum likelihood, including resistant a...
متن کامل